Goto

Collaborating Authors

 predict trustworthiness


Learning to Predict Trustworthiness with Steep Slope Loss

Neural Information Processing Systems

Understanding the trustworthiness of a prediction yielded by a classifier is critical for the safe and effective use of AI models. Prior efforts have been proven to be reliable on small-scale datasets. In this work, we study the problem of predicting trustworthiness on real-world large-scale datasets, where the task is more challenging due to high-dimensional features, diverse visual concepts, and a large number of samples. In such a setting, we observe that the trustworthiness predictors trained with prior-art loss functions, i.e., the cross entropy loss, focal loss, and true class probability confidence loss, are prone to view both correct predictions and incorrect predictions to be trustworthy.


Appendix for Learning to Predict Trustworthiness with Steep Slope Loss Y an Luo

Neural Information Processing Systems

By Hoeffding's bound, we have null The ViT (i.e., ViT Base/16) used in this work is implemented in the ASYML project The code is implemented in Python 3.8.5 with PyTorch 1.7.1 [ For the other experiments or analyses, we run one time. The implementation provides the pre-trained models on MNIST and CIFAR-10. License, while the implementation of ViT is licensed under the Apache-2.0 Ideally, we hope that all the confidences w.r.t. the positive class are on the right-hand side of the positive threshold while the ones w.r.t. the negative class are on the left-hand side of the negative The oracles that are used to generate the confidences are the ones used in Table 1. ImageNet validation set (stylized val) and the adversarial ImageNet validation set (adversarial val).


Learning to Predict Trustworthiness with Steep Slope Loss

Neural Information Processing Systems

Understanding the trustworthiness of a prediction yielded by a classifier is critical for the safe and effective use of AI models. Prior efforts have been proven to be reliable on small-scale datasets. In this work, we study the problem of predicting trustworthiness on real-world large-scale datasets, where the task is more challenging due to high-dimensional features, diverse visual concepts, and a large number of samples. In such a setting, we observe that the trustworthiness predictors trained with prior-art loss functions, i.e., the cross entropy loss, focal loss, and true class probability confidence loss, are prone to view both correct predictions and incorrect predictions to be trustworthy. Firstly, correct predictions are generally dominant over incorrect predictions.


Learning to Predict Trustworthiness with Steep Slope Loss

Neural Information Processing Systems

Understanding the trustworthiness of a prediction yielded by a classifier is critical for the safe and effective use of AI models. Prior efforts have been proven to be reliable on small-scale datasets. In this work, we study the problem of predicting trustworthiness on real-world large-scale datasets, where the task is more challenging due to high-dimensional features, diverse visual concepts, and a large number of samples. In such a setting, we observe that the trustworthiness predictors trained with prior-art loss functions, i.e., the cross entropy loss, focal loss, and true class probability confidence loss, are prone to view both correct predictions and incorrect predictions to be trustworthy. Firstly, correct predictions are generally dominant over incorrect predictions.